43 research outputs found

    Timing Tonogenesis: Evidence from Borrowing

    Get PDF
    Proceedings of the Twenty-Eighth Annual Meeting of the Berkeley Linguistics Society: Special Session on Tibeto-Burman and Southeast Asian Linguistics (2002

    Topic-Aware Response Generation in Task-Oriented Dialogue with Unstructured Knowledge Access

    Full text link
    To alleviate the problem of structured databases' limited coverage, recent task-oriented dialogue systems incorporate external unstructured knowledge to guide the generation of system responses. However, these usually use word or sentence level similarities to detect the relevant knowledge context, which only partially capture the topical level relevance. In this paper, we examine how to better integrate topical information in knowledge grounded task-oriented dialogue and propose ``Topic-Aware Response Generation'' (TARG), an end-to-end response generation model. TARG incorporates multiple topic-aware attention mechanisms to derive the importance weighting scheme over dialogue utterances and external knowledge sources towards a better understanding of the dialogue history. Experimental results indicate that TARG achieves state-of-the-art performance in knowledge selection and response generation, outperforming previous state-of-the-art by 3.2, 3.6, and 4.2 points in EM, F1 and BLEU-4 respectively on Doc2Dial, and performing comparably with previous work on DSTC9; both being knowledge-grounded task-oriented dialogue datasets.Comment: Findings of EMNLP 202

    In What Order Should Learners Learn Japanese Vocabulary? A Corpus-based Approach

    No full text
    This thesis attempts to answer the following two main research questions:1) In what order should learners of Japanese as a second language learn words and characters in order to be able to read Japanese? 2) How will the order vary according to the purpose of learning? To answer these questions, a Vocabulary Database for Reading Japanese (VDRJ) and a Character Database of Japanese (CDJ) were first developed from the Balanced Contemporary Corpus of Written Japanese (BCCWJ) 2009 monitor version (NINJAL, 2009) which contains book texts and internet-forum site texts with 33 million running words in total. Word and character rankings for international students, non-academic learners and general written Japanese were included in these databases. These rankings were proven to be valid for their respective purposes as they provided higher text coverage for the target texts than other texts. After analysing the use of vocabulary and characters in Japanese, three groups of domain-specific words, namely common academic words, limited-academic-domain words and literary words were extracted. In order to test the expected efficiency for learning these groups of words, an index entitled Text Covering Efficiency (TCE) in different types of texts was proposed. The TCE represents the expected return per unit of text length from learning a group of words. As such, the TCE score in the target text domain should determine the order in which words in this domain are most efficiently learned. Indeed, the extracted common academic words and limited-academic-domain words showed significantly higher text coverage and TCE scores in academic texts than in other texts. Literary words also provided high text coverage and high TCE scores in literary texts, despite a lower efficiency level than that of academic vocabulary in academic texts. Learning domain-specific words is expected to be much more efficient than learning other words at the intermediate level. At the advanced level or above, learning domain-specific words will be further more efficient in some domains such as the natural sciences. In sum, the TCE has been shown to provide useful information for deciding on the learning order of various groups of words. Other findings based on the analyses using the databases and word lists include the features of some indices for dispersion and adjusted frequency, lexical features of different media and genres, indexicality of the distributions of word origins and parts of speech, and the discrepancy between learning orders of words and Kanji. A Lexical Learning Possibility Index for a Reading Text (LEPIX) was also proposed for the simplification of a text as a vocabulary learning resource

    Different Durations of Diphthongs in Thai: A New Finding

    Get PDF
    Proceedings of the Twenty-Eighth Annual Meeting of the Berkeley Linguistics Society: Special Session on Tibeto-Burman and Southeast Asian Linguistics (2002

    Neural information extraction from natural language text

    Get PDF
    Natural language processing (NLP) deals with building computational techniques that allow computers to automatically analyze and meaningfully represent human language. With an exponential growth of data in this digital era, the advent of NLP-based systems has enabled us to easily access relevant information via a wide range of applications, such as web search engines, voice assistants, etc. To achieve it, a long-standing research for decades has been focusing on techniques at the intersection of NLP and machine learning. In recent years, deep learning techniques have exploited the expressive power of Artificial Neural Networks (ANNs) and achieved state-of-the-art performance in a wide range of NLP tasks. Being one of the vital properties, Deep Neural Networks (DNNs) can automatically extract complex features from the input data and thus, provide an alternative to the manual process of handcrafted feature engineering. Besides ANNs, Probabilistic Graphical Models (PGMs), a coupling of graph theory and probabilistic methods have the ability to describe causal structure between random variables of the system and capture a principled notion of uncertainty. Given the characteristics of DNNs and PGMs, they are advantageously combined to build powerful neural models in order to understand the underlying complexity of data. Traditional machine learning based NLP systems employed shallow computational methods (e.g., SVM or logistic regression) and relied on handcrafting features which is time-consuming, complex and often incomplete. However, deep learning and neural network based methods have recently shown superior results on various NLP tasks, such as machine translation, text classification, namedentity recognition, relation extraction, textual similarity, etc. These neural models can automatically extract an effective feature representation from training data. This dissertation focuses on two NLP tasks: relation extraction and topic modeling. The former aims at identifying semantic relationships between entities or nominals within a sentence or document. Successfully extracting the semantic relationships greatly contributes in building structured knowledge bases, useful in downstream NLP application areas of web search, question-answering, recommendation engines, etc. On other hand, the task of topic modeling aims at understanding the thematic structures underlying in a collection of documents. Topic modeling is a popular text-mining tool to automatically analyze a large collection of documents and understand topical semantics without actually reading them. In doing so, it generates word clusters (i.e., topics) and document representations useful in document understanding and information retrieval, respectively. Essentially, the tasks of relation extraction and topic modeling are built upon the quality of representations learned from text. In this dissertation, we have developed task-specific neural models for learning representations, coupled with relation extraction and topic modeling tasks in the realms of supervised and unsupervised machine learning paradigms, respectively. More specifically, we make the following contributions in developing neural models for NLP tasks: 1. Neural Relation Extraction: Firstly, we have proposed a novel recurrent neural network based architecture for table-filling in order to jointly perform entity and relation extraction within sentences. Then, we have further extended our scope of extracting relationships between entities across sentence boundaries, and presented a novel dependency-based neural network architecture. The two contributions lie in the supervised paradigm of machine learning. Moreover, we have contributed in building a robust relation extractor constrained by the lack of labeled data, where we have proposed a novel weakly-supervised bootstrapping technique. Given the contributions, we have further explored interpretability of the recurrent neural networks to explain their predictions for the relation extraction task. 2. Neural Topic Modeling: Besides the supervised neural architectures, we have also developed unsupervised neural models to learn meaningful document representations within topic modeling frameworks. Firstly, we have proposed a novel dynamic topic model that captures topics over time. Next, we have contributed in building static topic models without considering temporal dependencies, where we have presented neural topic modeling architectures that also exploit external knowledge, i.e., word embeddings to address data sparsity. Moreover, we have developed neural topic models that incorporate knowledge transfers using both the word embeddings and latent topics from many sources. Finally, we have shown improving neural topic modeling by introducing language structures (e.g., word ordering, local syntactic and semantic information, etc.) that deals with bag-of-words issues in traditional topic models. The class of proposed neural NLP models in this section are based on techniques at the intersection of PGMs, deep learning and ANNs. Here, the task of neural relation extraction employs neural networks to learn representations typically at the sentence level, without access to the broader document context. However, topic models have access to statistical information across documents. Therefore, we advantageously combine the two complementary learning paradigms in a neural composite model, consisting of a neural topic and a neural language model that enables us to jointly learn thematic structures in a document collection via the topic model, and word relations within a sentence via the language model. Overall, our research contributions in this dissertation extend NLP-based systems for relation extraction and topic modeling tasks with state-of-the-art performances
    corecore